Name:             loki-operator-controller-manager-6bf69c47b7-8fjw7
Namespace:        openshift-operators-redhat
Priority:         0
Service Account:  loki-operator-controller-manager
Node:             crc/192.168.126.11
Start Time:       Tue, 09 Dec 2025 09:59:03 +0000
Labels:           app.kubernetes.io/managed-by=operator-lifecycle-manager
                  app.kubernetes.io/name=loki-operator
                  app.kubernetes.io/part-of=cluster-logging
                  name=loki-operator-controller-manager
                  pod-template-hash=6bf69c47b7
Annotations:      alm-examples:
                    [
                      {
                        "apiVersion": "loki.grafana.com/v1",
                        "kind": "AlertingRule",
                        "metadata": {
                          "name": "alertingrule-sample"
                        },
                        "spec": {
                          "groups": [
                            {
                              "interval": "10m",
                              "name": "alerting-rules-group",
                              "rules": [
                                {
                                  "alert": "HighPercentageError",
                                  "annotations": {
                                    "summary": "High request latency"
                                  },
                                  "expr": "sum(rate({app=\"foo\", env=\"production\"} |= \"error\" [5m])) by (job)\n  /\nsum(rate({app=\"foo\", env=\"producti...
                                  "for": "10m",
                                  "labels": {
                                    "severity": "page"
                                  }
                                },
                                {
                                  "alert": "HttpCredentialsLeaked",
                                  "annotations": {
                                    "message": "{{ $labels.job }} is leaking http basic auth credentials."
                                  },
                                  "expr": "sum by (cluster, job, pod) (count_over_time({namespace=\"prod\"} |~ \"http(s?)://(\\\\w+):(\\\\w+)@\" [5m]) \u003e ...
                                  "for": "10m",
                                  "labels": {
                                    "severity": "critical"
                                  }
                                }
                              ]
                            }
                          ],
                          "tenantID": "test-tenant"
                        }
                      },
                      {
                        "apiVersion": "loki.grafana.com/v1",
                        "kind": "LokiStack",
                        "metadata": {
                          "name": "lokistack-sample"
                        },
                        "spec": {
                          "size": "1x.small",
                          "storage": {
                            "secret": {
                              "name": "test"
                            }
                          },
                          "storageClassName": "standard"
                        }
                      },
                      {
                        "apiVersion": "loki.grafana.com/v1",
                        "kind": "RecordingRule",
                        "metadata": {
                          "name": "recordingrule-sample"
                        },
                        "spec": {
                          "groups": [
                            {
                              "interval": "10m",
                              "name": "recording-rules-group",
                              "rules": [
                                {
                                  "expr": "sum(rate({container=\"myservice\"}[10m]))\n",
                                  "record": "myservice:requests:rate10m"
                                },
                                {
                                  "expr": "sum(rate({container=\"otherservice\"}[1m]))\n",
                                  "record": "otherservice:requests:rate1m"
                                }
                              ]
                            }
                          ],
                          "tenantID": "test-tenant"
                        }
                      },
                      {
                        "apiVersion": "loki.grafana.com/v1",
                        "kind": "RulerConfig",
                        "metadata": {
                          "name": "rulerconfig-sample"
                        },
                        "spec": {
                          "alertmanager": {
                            "discovery": {
                              "enableSRV": true,
                              "refreshInterval": "1m"
                            },
                            "enableV2": true,
                            "endpoints": [
                              "http://alertmanager-host1.mycompany.org",
                              "http://alertmanager-host2.mycompany.org"
                            ],
                            "externalLabels": {
                              "environment": "production",
                              "region": "us-east-2"
                            },
                            "externalUrl": "http://www.mycompany.org/alerts",
                            "notificationQueue": {
                              "capacity": 1000,
                              "forGracePeriod": "10m",
                              "forOutageTolerance": "1h",
                              "resendDelay": "1m",
                              "timeout": "30s"
                            }
                          },
                          "evaluationInterval": "1m",
                          "pollInterval": "1m",
                          "remoteWrite": {
                            "client": {
                              "authorization": "basic",
                              "authorizationSecretName": "my-secret-resource",
                              "name": "remote-write-log-metrics",
                              "proxyUrl": "http://proxy-host.mycompany.org",
                              "relabelConfigs": [
                                {
                                  "action": "replace",
                                  "regex": "ALERTS.*",
                                  "replacement": "$1",
                                  "separator": "",
                                  "sourceLabels": [
                                    "labelc",
                                    "labeld"
                                  ],
                                  "targetLabel": "labelnew"
                                }
                              ],
                              "timeout": "30s",
                              "url": "http://remote-write-host.mycompany.org"
                            },
                            "enabled": true,
                            "refreshPeriod": "10s"
                          }
                        }
                      }
                    ]
                  capabilities: Full Lifecycle
                  categories: OpenShift Optional, Logging & Tracing
                  certified: false
                  com.redhat.delivery.appregistry: false
                  com.redhat.delivery.operator.bundle: true
                  com.redhat.openshift.versions: v4.16-v4.20
                  containerImage:
                    registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:ce1c37005ef2613c22343b3841f495597103f0831839d017f8a9dad8a00275c6
                  createdAt: 2025-02-18T18:01:39Z
                  description:
                    The Loki Operator for OCP provides a means for configuring and managing a Loki stack for cluster logging.
                    ## Prerequisites and Requirements
                    ### Loki Operator Namespace
                    The Loki Operator must be deployed to the global operator group namespace `openshift-operators-redhat`.
                    ### Memory Considerations
                    Loki is a memory intensive application.  The initial
                    set of OCP nodes may not be large enough to support the Loki stack.  Additional OCP nodes must be added
                    to the OCP cluster if you desire to run with the recommended (or better) memory.
                  features.operators.openshift.io/cnf: false
                  features.operators.openshift.io/cni: false
                  features.operators.openshift.io/csi: false
                  features.operators.openshift.io/disconnected: true
                  features.operators.openshift.io/fips-compliant: true
                  features.operators.openshift.io/proxy-aware: true
                  features.operators.openshift.io/tls-profiles: true
                  features.operators.openshift.io/token-auth-aws: true
                  features.operators.openshift.io/token-auth-azure: true
                  features.operators.openshift.io/token-auth-gcp: true
                  k8s.ovn.org/pod-networks:
                    {"default":{"ip_addresses":["10.217.0.47/23"],"mac_address":"0a:58:0a:d9:00:2f","gateway_ips":["10.217.0.1"],"routes":[{"dest":"10.217.0.0...
                  k8s.v1.cni.cncf.io/network-status:
                    [{
                        "name": "ovn-kubernetes",
                        "interface": "eth0",
                        "ips": [
                            "10.217.0.47"
                        ],
                        "mac": "0a:58:0a:d9:00:2f",
                        "default": true,
                        "dns": {}
                    }]
                  kubectl.kubernetes.io/default-container: manager
                  olm.operatorGroup: loki-operator
                  olm.operatorNamespace: openshift-operators-redhat
                  olm.skipRange: >=6.0.0-0 <6.2.6
                  olm.targetNamespaces: 
                  olmcahash: 7257a5784b5e82d6b7e3e3e9e33ca5c5d7a8a942176a22c2314299cb6f54b33e
                  openshift.io/scc: restricted-v2
                  operatorframework.io/cluster-monitoring: true
                  operatorframework.io/properties:
                    {"properties":[{"type":"olm.gvk","value":{"group":"loki.grafana.com","kind":"AlertingRule","version":"v1"}},{"type":"olm.gvk","value":{"gr...
                  operatorframework.io/suggested-namespace: openshift-operators-redhat
                  operators.openshift.io/valid-subscription: ["OpenShift Container Platform", "OpenShift Platform Plus"]
                  operators.operatorframework.io.bundle.channel.default.v1: stable-6.2
                  operators.operatorframework.io.bundle.channels.v1: stable-6.2
                  operators.operatorframework.io.bundle.manifests.v1: manifests/
                  operators.operatorframework.io.bundle.mediatype.v1: registry+v1
                  operators.operatorframework.io.bundle.metadata.v1: metadata/
                  operators.operatorframework.io.bundle.package.v1: loki-operator
                  operators.operatorframework.io.metrics.mediatype.v1: metrics+v1
                  operators.operatorframework.io.metrics.project_layout: go.kubebuilder.io/v4
                  operators.operatorframework.io.test.config.v1: tests/scorecard/
                  operators.operatorframework.io.test.mediatype.v1: scorecard+v1
                  operators.operatorframework.io/builder: operator-sdk-unknown
                  operators.operatorframework.io/project_layout: go.kubebuilder.io/v4
                  seccomp.security.alpha.kubernetes.io/pod: runtime/default
                  support: AOS Cluster Logging
Status:           Running
SeccompProfile:   RuntimeDefault
IP:               10.217.0.47
IPs:
  IP:           10.217.0.47
Controlled By:  ReplicaSet/loki-operator-controller-manager-6bf69c47b7
Containers:
  manager:
    Container ID:  cri-o://8bc19e9a1e4ac657a0eb38e7d4c3df469bfc33710342fbae996c94d3caf8fe41
    Image:         registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:ce1c37005ef2613c22343b3841f495597103f0831839d017f8a9dad8a00275c6
    Image ID:      registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:ce1c37005ef2613c22343b3841f495597103f0831839d017f8a9dad8a00275c6
    Ports:         9443/TCP, 8080/TCP
    Host Ports:    0/TCP, 0/TCP
    Command:
      /manager
    Args:
      --config=controller_manager_config.yaml
    State:          Running
      Started:      Tue, 09 Dec 2025 11:14:25 +0000
    Last State:     Terminated
      Reason:       Error
      Exit Code:    1
      Started:      Tue, 09 Dec 2025 09:59:14 +0000
      Finished:     Tue, 09 Dec 2025 11:14:22 +0000
    Ready:          True
    Restart Count:  1
    Liveness:       http-get http://:8081/healthz delay=15s timeout=1s period=20s #success=1 #failure=3
    Readiness:      http-get http://:8081/readyz delay=5s timeout=1s period=10s #success=1 #failure=3
    Environment:
      RELATED_IMAGE_LOKI:       registry.redhat.io/openshift-logging/logging-loki-rhel9@sha256:1bd60df77d8be8eae3551f68a3a55a464610be839b0c0556600c7f1a36887919
      RELATED_IMAGE_GATEWAY:    registry.redhat.io/openshift-logging/lokistack-gateway-rhel9@sha256:ad8b5cee31875481991739c882fd765f01d0391184de7e2e19e8394efda5a4d6
      RELATED_IMAGE_OPA:        registry.redhat.io/openshift-logging/opa-openshift-rhel9@sha256:62c2e9714d7de73861ca4a607a9509ddc9771399569015a5ad8f8b42639ebaae
      OPERATOR_CONDITION_NAME:  loki-operator.v6.2.6
    Mounts:
      /apiserver.local.config/certificates from apiservice-cert (rw)
      /controller_manager_config.yaml from manager-config (rw,path="controller_manager_config.yaml")
      /tmp/k8s-webhook-server/serving-certs from webhook-cert (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-s549x (ro)
  kube-rbac-proxy:
    Container ID:  cri-o://14bd9def8a01c7ee609b2e20c0cd5f94e258b066263ccab45319d31e259ad029
    Image:         registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
    Image ID:      registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24
    Port:          8443/TCP
    Host Port:     0/TCP
    Args:
      --secure-listen-address=0.0.0.0:8443
      --upstream=http://127.0.0.1:8080/
      --logtostderr=true
      --tls-cert-file=/var/run/secrets/serving-cert/tls.crt
      --tls-private-key-file=/var/run/secrets/serving-cert/tls.key
      --tls-cipher-suites=TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256,TLS_RSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_256_GCM_SHA384,TLS_RSA_WITH_AES_128_CBC_SHA256
      --tls-min-version=VersionTLS12
      --v=0
    State:          Running
      Started:      Tue, 09 Dec 2025 09:59:48 +0000
    Ready:          True
    Restart Count:  0
    Environment:
      OPERATOR_CONDITION_NAME:  loki-operator.v6.2.6
    Mounts:
      /apiserver.local.config/certificates from apiservice-cert (rw)
      /tmp/k8s-webhook-server/serving-certs from webhook-cert (rw)
      /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-s549x (ro)
      /var/run/secrets/serving-cert from loki-operator-metrics-cert (rw)
Conditions:
  Type                        Status
  PodReadyToStartContainers   True 
  Initialized                 True 
  Ready                       True 
  ContainersReady             True 
  PodScheduled                True 
Volumes:
  manager-config:
    Type:      ConfigMap (a volume populated by a ConfigMap)
    Name:      loki-operator-manager-config
    Optional:  false
  loki-operator-metrics-cert:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  loki-operator-metrics
    Optional:    true
  apiservice-cert:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  loki-operator-controller-manager-service-cert
    Optional:    false
  webhook-cert:
    Type:        Secret (a volume populated by a Secret)
    SecretName:  loki-operator-controller-manager-service-cert
    Optional:    false
  kube-api-access-s549x:
    Type:                    Projected (a volume that contains injected data from multiple sources)
    TokenExpirationSeconds:  3607
    ConfigMapName:           kube-root-ca.crt
    ConfigMapOptional:       <nil>
    DownwardAPI:             true
    ConfigMapName:           openshift-service-ca.crt
    ConfigMapOptional:       <nil>
QoS Class:                   BestEffort
Node-Selectors:              kubernetes.io/os=linux
Tolerations:                 node.kubernetes.io/not-ready:NoExecute op=Exists for 300s
                             node.kubernetes.io/unreachable:NoExecute op=Exists for 300s
Events:
  Type     Reason          Age                    From               Message
  ----     ------          ----                   ----               -------
  Normal   Scheduled       80m                    default-scheduler  Successfully assigned openshift-operators-redhat/loki-operator-controller-manager-6bf69c47b7-8fjw7 to crc
  Normal   AddedInterface  80m                    multus             Add eth0 [10.217.0.47/23] from ovn-kubernetes
  Normal   Pulling         80m                    kubelet            Pulling image "registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:ce1c37005ef2613c22343b3841f495597103f0831839d017f8a9dad8a00275c6"
  Normal   Pulled          80m                    kubelet            Successfully pulled image "registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:ce1c37005ef2613c22343b3841f495597103f0831839d017f8a9dad8a00275c6" in 9.489s (9.489s including waiting). Image size: 181660366 bytes.
  Warning  Failed          80m                    kubelet            Failed to pull image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24": rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled
  Warning  Failed          80m                    kubelet            Error: ErrImagePull
  Normal   BackOff         80m (x3 over 80m)      kubelet            Back-off pulling image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24"
  Warning  Failed          80m (x3 over 80m)      kubelet            Error: ImagePullBackOff
  Normal   Pulling         80m (x2 over 80m)      kubelet            Pulling image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24"
  Normal   Created         80m                    kubelet            Created container kube-rbac-proxy
  Normal   Pulled          80m                    kubelet            Successfully pulled image "registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:145e9784b681ac7defa0a1547c03a6db9a587bf9be2820428eee84f58f8f1f24" in 1.537s (1.537s including waiting). Image size: 461588476 bytes.
  Normal   Started         80m                    kubelet            Started container kube-rbac-proxy
  Warning  ProbeError      5m35s (x2 over 5m56s)  kubelet            Readiness probe error: Get "http://10.217.0.47:8081/readyz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
body:
  Warning  Unhealthy  5m35s (x2 over 5m56s)  kubelet  Readiness probe failed: Get "http://10.217.0.47:8081/readyz": context deadline exceeded (Client.Timeout exceeded while awaiting headers)
  Normal   Pulled     5m27s                  kubelet  Container image "registry.redhat.io/openshift-logging/loki-rhel9-operator@sha256:ce1c37005ef2613c22343b3841f495597103f0831839d017f8a9dad8a00275c6" already present on machine
  Normal   Started    5m25s (x2 over 80m)    kubelet  Started container manager
  Normal   Created    5m25s (x2 over 80m)    kubelet  Created container manager
