Name:             loki-operator-controller-manager-775d8c8b9c-rkqj2
Namespace:        openshift-operators-redhat
Priority:         0
Service Account:  loki-operator-controller-manager
Node:             crc/192.168.126.11
Start Time:       Fri, 23 Jan 2026 11:02:01 +0000
Labels:           app.kubernetes.io/managed-by=operator-lifecycle-manager
                  app.kubernetes.io/name=loki-operator
                  app.kubernetes.io/part-of=cluster-logging
                  name=loki-operator-controller-manager
                  pod-template-hash=775d8c8b9c
Annotations:      alm-examples:
                    [
                      {
                        "apiVersion": "loki.grafana.com/v1",
                        "kind": "AlertingRule",
                        "metadata": {
                          "name": "alertingrule-sample"
                        },
                        "spec": {
                          "groups": [
                            {
                              "interval": "10m",
                              "name": "alerting-rules-group",
                              "rules": [
                                {
                                  "alert": "HighPercentageError",
                                  "annotations": {
                                    "summary": "High request latency"
                                  },
                                  "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"producti...
                                  "for": "10m",
                                  "labels": {
                                    "severity": "page"
                                  }
                                },
                                {
                                  "alert": "HttpCredentialsLeaked",
                                  "annotations": {
                                    "message": "{{ $labels.job }} is leaking http basic auth credentials."
                                  },
                                  "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e ...
                                  "for": "10m",
                                  "labels": {
                                    "severity": "critical"
                                  }
                                }
                              ]
                            }
                          ],
                          "tenantID": "test-tenant"
                        }
                      },
                      {
                        "apiVersion": "loki.grafana.com/v1",
                        "kind": "LokiStack",
                        "metadata": {
                          "name": "lokistack-sample"
                        },
                        "spec": {
                          "size": "1x.small",
                          "storage": {
                            "secret": {
                              "name": "test"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      {
                        "apiVersion": "loki.grafana.com/v1",
                        "kind": "RecordingRule",
                        "metadata": {
                          "name": "recordingrule-sample"
                        },
                        "spec": {
                          "groups": [
                            {
                              "interval": "10m",
                              "name": "recording-rules-group",
                              "rules": [
                                {
                                  "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                                  "record": "myservice:requests:rate10m"
                                },
                                {
                                  "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                                  "record": "otherservice:requests:rate1m"
                                }
                              ]
                            }
                          ],
                          "tenantID": "test-tenant"
                        }
                      },
                      {
                        "apiVersion": "loki.grafana.com/v1",
                        "kind": "RulerConfig",
                        "metadata": {
                          "name": "rulerconfig-sample"
                        },
                        "spec": {
                          "alertmanager": {
                            "discovery": {
                              "enableSRV": true,
                              "refreshInterval": "1m"
                            },
                            "enableV2": true,
                            "endpoints": [
                              "http://alertmanager-host1.mycompany.org",
                              "http://alertmanager-host2.mycompany.org"
                            ],
                            "externalLabels": {
                              "environment": "production",
                              "region": "us-east-2"
                            },
                            "externalUrl": "http://www.mycompany.org/alerts",
                            "notificationQueue": {
                              "capacity": 1000,
                              "forGracePeriod": "10m",
                              "forOutageTolerance": "1h",
                              "resendDelay": "1m",
                              "timeout": "30s"
                            }
                          },
                          "evaluationInterval": "1m",
                          "pollInterval": "1m",
                          "remoteWrite": {
                            "client": {
                              "authorization": "basic",
                              "authorizationSecretName": "my-secret-resource",
                              "name": "remote-write-log-metrics",
                              "proxyUrl": "http://proxy-host.mycompany.org",
                              "relabelConfigs": [
                                {
                                  "action": "replace",
                                  "regex": "ALERTS.*",
                                  "replacement": "$1",
                                  "separator": "",
                                  "sourceLabels": [
                                    "labelc",
                                    "labeld"
                                  ],
                                  "targetLabel": "labelnew"
                                }
                              ],
                              "timeout": "30s",
                              "url": "http://remote-write-host.mycompany.org"
                            },
                            "enabled": true,
                            "refreshPeriod": "10s"
                          }
                        }
                      }
                    ]
                  capabilities: Full Lifecycle
                  categories: OpenShift Optional, Logging & Tracing
                  certified: false
                  com.redhat.delivery.appregistry: false
                  com.redhat.delivery.operator.bundle: true
                  com.redhat.openshift.versions: v4.16-v4.20
                  containerImage:
                    registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f7fde57f19ef352510cfca9f073ab4f73cd769bd097e1b1d771070d58f682240
                  createdAt: 2025-03-06T11:32:52Z
                  description:
                    The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
                    ## Prerequisites and Requirements
                    ### Loki Operator Namespace
                    The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
                    ### Memory Considerations
                    Loki is a memory intensive application.  The initial
                    set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
                    to the OCP cluster if you desire to run with the recommended (or better) memory.
                  features.operators.openshift.io/cnf: false
                  features.operators.openshift.io/cni: false
                  features.operators.openshift.io/csi: false
                  features.operators.openshift.io/disconnected: true
                  features.operators.openshift.io/fips-compliant: true
                  features.operators.openshift.io/proxy-aware: true
                  features.operators.openshift.io/tls-profiles: true
                  features.operators.openshift.io/token-auth-aws: true
                  features.operators.openshift.io/token-auth-azure: true
                  features.operators.openshift.io/token-auth-gcp: true
                  k8s.ovn.org/pod-networks:
                    {"default":{"ip_addresses":["10.217.0.49/23"],"mac_address":"0a:58:0a:d9:00:31","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0.0...
                  k8s.v1.cni.cncf.io/network-status:
                    [{
                        "name": "ovn-kubernetes",
                        "interface": "eth0",
                        "ips": [
                            "10.217.0.49"
                        ],
                        "mac": "0a:58:0a:d9:00:31",
                        "default": true,
                        "dns": {}
                    }]
                  kubectl.kubernetes.io/default-container: manager
                  olm.operatorGroup: loki-operator
                  olm.operatorNamespace: openshift-operators-redhat
                  olm.skipRange: >=6.0.0-0 <6.2.7
                  olm.targetNamespaces: 
                  olmcahash: cf591960c3f41380dddc1d2e716a7206be270d9b9269c4e29d5d3a2482a69e2b
                  openshift.io/scc: restricted-v2
                  operatorframework.io/cluster-monitoring: true
                  operatorframework.io/properties:
                    {"properties":[{"type":"olm.gvk","value":{"group":"loki.grafana.com","kind":"AlertingRule","version":"v1"}},{"type":"olm.gvk","value":{"gr...
                  operatorframework.io/suggested-namespace: openshift-operators-redhat
                  operators.openshift.io/valid-subscription: ["OpenShift Container Platform", "OpenShift Platform Plus"]
                  operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
                  operators.operatorframework.io.bundle.channels.v1: stable-6.2
                  operators.operatorframework.io.bundle.manifests.v1: manifests/
                  operators.operatorframework.io.bundle.mediatype.v1: registry+v1
                  operators.operatorframework.io.bundle.metadata.v1: metadata/
                  operators.operatorframework.io.bundle.package.v1: loki-operator
                  operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
                  operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
                  operators.operatorframework.io.test.config.v1: tests/scorecard/
                  operators.operatorframework.io.test.mediatype.v1: scorecard+v1
                  operators.operatorframework.io/builder: operator-sdk-unknown
                  operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
                  seccomp.security.alpha.kubernetes.io/pod: runtime/default
                  support: AOS Cluster Logging
Status:           Running
SeccompProfile:   RuntimeDefault
IP:               10.217.0.49
IPs:
  IP:           10.217.0.49
Controlled By:  ReplicaSet/loki-operator-controller-manager-775d8c8b9c
Containers:
  manager:
    Container ID:  cri-o://24e9a132da8f0396b3fe43ad641387517b800dadcb1f8896e78a57827c814679
    Image:         registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f7fde57f19ef352510cfca9f073ab4f73cd769bd097e1b1d771070d58f682240
    Image ID:      registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:c88d0b886366991bb212aa43f5c983f284f8ebeb9da44c64f08324c2ba2d999d
    Ports:         9443/TCP, 8080/TCP
    Host Ports:    0/TCP, 0/TCP
    Command:
      /manager
    Args:
      --config=controller_manager_config.yaml
    State:          Running
      Started:      Fri, 23 Jan 2026 12:13:44 +0000
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Fri, 23 Jan 2026 11:02:09 +0000
      Finished:     Fri, 23 Jan 2026 12:13:40 +0000
    Ready:          True
    Restart Count:  1
    Liveness:       http-get http://:8081/healthz delay=15s timeout=1s period=20s #success=1 #failure=3
    Readiness:      http-get http://:8081/readyz delay=5s timeout=1s period=10s #success=1 #failure=3
    Environment:
      RELATED_IMAGE_LOKI:       registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:8c073f8de8288bf63489a400d7f93e96562652832965d33e0515f48d46282459
      RELATED_IMAGE_GATEWAY:    registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:574afd46f23637b1685ddb00b647bb71d2e2d527611182fd2826bb818c5eb198
      RELATED_IMAGE_OPA:        registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:182dc5ab627d64b5e565a2edb170903ac381fc534b4366fc5ffae088ea0de5e5
      OPERATOR_CONDITION_NAME:  loki-operator.v6.2.7
    Mounts:
      /apiserver.local.config/certificates from apiservice-cert (rw)
      /controller_manager_config.yaml from manager-config (rw,path="controller_manager_config.yaml")
      /tmp/k8s-webhook-server/serving-certs from webhook-cert (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-cmh98 (ro)
  kube-rbac-proxy:
    Container ID:  cri-o://01652aa388db634d40edffb1d921d6d4a149eb7b5692ceca2006b4a5d8fb6c7b
    Image:         registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
    Image ID:      registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
    Port:          8443/TCP
    Host Port:     0/TCP
    Args:
      --secure-listen-address=0.0.0.0:8443
      --upstream=http://127.0.0.1:8080/
      --logtostderr=true
      --tls-cert-file=/var/run/secrets/serving-cert/tls.crt
      --tls-private-key-file=/var/run/secrets/serving-cert/tls.key
      --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256,TLS_RSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_256_GCM_SHA384,TLS_RSA_WITH_AES_128_CBC_SHA256
      --tls-min-version=VersionTLS12
      --v=0
    State:          Running
      Started:      Fri, 23 Jan 2026 11:02:20 +0000
    Ready:          True
    Restart Count:  0
    Environment:
      OPERATOR_CONDITION_NAME:  loki-operator.v6.2.7
    Mounts:
      /apiserver.local.config/certificates from apiservice-cert (rw)
      /tmp/k8s-webhook-server/serving-certs from webhook-cert (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-cmh98 (ro)
      /var/run/secrets/serving-cert from loki-operator-metrics-cert (rw)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  manager-config:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      loki-operator-manager-config
    Optional:  false
  loki-operator-metrics-cert:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  loki-operator-metrics
    Optional:    true
  apiservice-cert:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  loki-operator-controller-manager-service-cert
    Optional:    false
  webhook-cert:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  loki-operator-controller-manager-service-cert
    Optional:    false
  kube-api-access-cmh98:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   BestEffort
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason          Age    From               Message
  ----     ------          ----   ----               -------
  Normal   Scheduled       77m    default-scheduler  Successfully assigned openshift-operators-redhat/loki-operator-controller-manager-775d8c8b9c-rkqj2 to crc
  Normal   AddedInterface  77m    multus             Add eth0 [10.217.0.49/23] from ovn-kubernetes
  Normal   Pulling         77m    kubelet            Pulling image "registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f7fde57f19ef352510cfca9f073ab4f73cd769bd097e1b1d771070d58f682240"
  Normal   Pulled          77m    kubelet            Successfully pulled image "registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f7fde57f19ef352510cfca9f073ab4f73cd769bd097e1b1d771070d58f682240" in 7.251s (7.251s including waiting). Image size: 186857893 bytes.
  Normal   Pulling         77m    kubelet            Pulling image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24"
  Normal   Pulled          77m    kubelet            Successfully pulled image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24" in 10.12s (10.12s including waiting). Image size: 461588476 bytes.
  Normal   Created         77m    kubelet            Created container kube-rbac-proxy
  Normal   Started         77m    kubelet            Started container kube-rbac-proxy
  Warning  ProbeError      6m34s  kubelet            Liveness probe error: Get "http://10.217.0.49:8081/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
body:
  Warning  Unhealthy   6m34s                  kubelet  Liveness probe failed: Get "http://10.217.0.49:8081/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Warning  ProbeError  6m22s (x4 over 6m43s)  kubelet  Readiness probe error: Get "http://10.217.0.49:8081/readyz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
body:
  Warning  Unhealthy   6m22s (x4 over 6m43s)  kubelet  Readiness probe failed: Get "http://10.217.0.49:8081/readyz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Warning  ProbeError  6m15s                  kubelet  Liveness probe error: Get "http://10.217.0.49:8081/healthz": dial tcp 10.217.0.49:8081: connect: connection refused
body:
  Warning  Unhealthy   6m15s  kubelet  Liveness probe failed: Get "http://10.217.0.49:8081/healthz": dial tcp 10.217.0.49:8081: connect: connection refused
  Warning  ProbeError  6m15s  kubelet  Readiness probe error: Get "http://10.217.0.49:8081/readyz": dial tcp 10.217.0.49:8081: connect: connection refused
body:
  Warning  Unhealthy  6m15s                kubelet  Readiness probe failed: Get "http://10.217.0.49:8081/readyz": dial tcp 10.217.0.49:8081: connect: connection refused
  Normal   Pulled     6m13s                kubelet  Container image "registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:f7fde57f19ef352510cfca9f073ab4f73cd769bd097e1b1d771070d58f682240" already present on machine
  Normal   Started    6m12s (x2 over 77m)  kubelet  Started container manager
  Normal   Created    6m12s (x2 over 77m)  kubelet  Created container manager
